Sampling Bounds for Stochastic Optimization

نویسندگان

  • Moses Charikar
  • Chandra Chekuri
  • Martin Pál
چکیده

A large class of stochastic optimization problems can be modeled as minimizing an objective function f that depends on a choice of a vector x ∈ X, as well as on a random external parameter ω ∈ Ω given by a probability distribution π. The value of the objective function is a random variable and often the goal is to find an x ∈ X to minimize the expected cost Eω[fω(x)]. Each ω is referred to as a scenario. We consider the case when Ω is large or infinite and we are allowed to sample from π in a black-box fashion. A common method, known as the SAA method (sample average approximation), is to pick sufficiently many independent samples from π and use them to approximate π and correspondingly Eω[fω(x)]. This is one of several scenario reduction methods used in practice. There has been substantial recent interest in two-stage stochastic versions of combinatorial optimization problems which can be modeled by the framework described above. In particular, we are interested in the model where a parameter λ bounds the relative factor by which costs increase if decisions are delayed to the second stage. Although the SAA method has been widely analyzed, the known bounds on the number of samples required for a (1 + ε) approximation depend on the variance of π even when λ is assumed to be a fixed constant. Shmoys and Swamy [13, 14] proved that a polynomial number of samples suffice when f can be modeled as a linear or convex program. They used modifications to the ellipsoid method to prove this. In this paper we give a different proof, based on earlier methods of Kleywegt, Shapiro, Homem-De-Mello [6] and others, that a polynomial number of samples suffice for the SAA method. Our proof is not based on computational properties of f and hence also applies to integer programs. We further show that small variations of the SAA method suffice to obtain a bound on the sample size even when we have only an approximation algorithm to solve the sampled problem. We are thus able to extend a number of algorithms designed for the case when π is given explicitly to the case when π is given as a black-box sampling oracle. ? Supported by NSF ITR grant CCR-0205594, DOE Early Career Principal Investigator award DE-FG02-02ER25540, NSF CAREER award CCR-0237113, an Alfred P. Sloan Fellowship and a Howard B. Wentz Jr. Junior Faculty Award. [email protected] ?? Supported in part by an ONR basic research grant MA14681000 to Lucent Bell Labs. [email protected] ? ? ? Work done while at DIMACS, supported by NSF grant EIA 02-05116. [email protected]

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Scenario Set Partition Dual Bounds for Multistage Stochastic Programming: A Hierarchy of Bounds and a Partition Sampling Approach

We consider multistage stochastic programming problems in which the random parameters have finite support, leading to optimization over a finite scenario set. We propose a hierarchy of bounds based on partitions of the scenario set into subsets of (nearly) equal cardinality. These expected partition (EP) bounds coincide with EGSO bounds provided by Sandıkçı et al. (2013) in the case that the su...

متن کامل

A stochastic version analysis of an M/G/1 retrial queue with Bernoulli‎ ‎schedule‎

‎In this work‎, ‎we derive insensitive bounds for various performance measures of a single-server‎ ‎retrial queue with generally distributed inter-retrial times and Bernoulli schedule‎, ‎under the special‎ ‎assumption that only the customer at the head of the orbit queue (i.e.‎, ‎a‎ ‎FCFS discipline governing the flow from the orbit to the server) is allowed‎ ‎to occupy the server‎. ‎The method...

متن کامل

Faster Optimization through Adaptive Importance Sampling

The current state of the art stochastic optimization algorithms (SGD, SVRG, SCD, SDCA, etc.) are based on sampling one active datapoint uniformly at random in each iteration. Changing these probabilities to better reflect the importance of each datapoint is a natural and powerful idea. In this thesis we analyze Stochastic Coordinate Descent methods with fixed non-uniform and adaptive sampling. ...

متن کامل

Assessing Solution Quality in Stochastic Programs via Sampling

Determining if a solution is optimal or near optimal is fundamental in optimization theory, algorithms, and computation. For instance, Karush-Kuhn-Tucker conditions provide necessary and sufficient optimality conditions for certain classes of problems, and bounds on optimality gaps are frequently used as part of optimization algorithms. Such bounds are obtained through Lagrangian, integrality, ...

متن کامل

Local Smoothness in Variance Reduced Optimization

We propose a family of non-uniform sampling strategies to provably speed up a class of stochastic optimization algorithms with linear convergence including Stochastic Variance Reduced Gradient (SVRG) and Stochastic Dual Coordinate Ascent (SDCA). For a large family of penalized empirical risk minimization problems, our methods exploit data dependent local smoothness of the loss functions near th...

متن کامل

Stochastic bounds for a single server queue with general retrial times

We propose to use a mathematical method based on stochastic comparisons of Markov chains in order to derive performance indice bounds‎. ‎The main goal of this paper is to investigate various monotonicity properties of a single server retrial queue with first-come-first-served (FCFS) orbit and general retrial times using the stochastic ordering techniques‎.

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2005